# MLX optimization

Diffucoder 7B Cpgrpo 8bit
DiffuCoder-7B-cpGRPO-8bit is a code generation model converted to MLX format, based on apple/DiffuCoder-7B-cpGRPO, and is specifically designed to provide developers with an efficient code generation tool.
Large Language Model Other
D
mlx-community
272
2
ERNIE 4.5 21B A3B PT 8bit
Apache-2.0
ERNIE-4.5-21B-A3B-PT-8bit is an 8-bit quantized version of Baidu's ERNIE-4.5-21B-A3B-PT model, converted to MLX format and suitable for Apple Silicon devices.
Large Language Model Supports Multiple Languages
E
mlx-community
123
1
Diffucoder 7B Cpgrpo 4bit
DiffuCoder-7B-cpGRPO-4bit is a 4-bit quantized version converted from the Apple DiffuCoder-7B-cpGRPO model, optimized for the MLX framework.
Large Language Model Other
D
mlx-community
218
1
Hunyuan A13B Instruct 5bit
Other
A 5-bit quantized version converted from Tencent's Hunyuan A13B large model, suitable for the MLX framework on Apple chips
Large Language Model
H
mlx-community
169
2
Hunyuan A13B Instruct 4bit
Other
The 4-bit quantization version of Tencent Hunyuan A13B large language model, suitable for instruction following tasks
Large Language Model
H
mlx-community
201
4
Jan Nano 8bit
Apache-2.0
Jan-nano-8bit is an 8-bit quantized version converted from the Menlo/Jan-nano model, optimized for the MLX framework and suitable for text generation tasks.
Large Language Model
J
mlx-community
188
1
Qwen3 235B A22B 4bit DWQ 053125
Apache-2.0
This is a 4-bit quantized version converted from the Qwen3-235B-A22B-8bit model, optimized for the MLX framework and suitable for text generation tasks.
Large Language Model
Q
mlx-community
200
1
Josiefied DeepSeek R1 0528 Qwen3 8B Abliterated V1 8bit
This is an 8-bit quantized version in MLX format converted from the DeepSeek-R1-0528-Qwen3-8B model, suitable for text generation tasks.
Large Language Model
J
mlx-community
847
1
Llama 3.3 70B Instruct 4bit DWQ
4-bit DWQ quantized version of the Llama 3.3 70B instruction-tuned model, optimized for efficient inference on the MLX framework
Large Language Model Supports Multiple Languages
L
mlx-community
140
2
Devstral Small 2505 4bit DWQ
Apache-2.0
This is a 4-bit quantized language model in MLX format, suitable for text generation tasks.
Large Language Model Supports Multiple Languages
D
mlx-community
238
3
Devstral Small 2505 8bit
Apache-2.0
Devstral-Small-2505-8bit is an 8-bit quantized model converted from mistralai/Devstral-Small-2505, suitable for the MLX framework and supporting text generation tasks in multiple languages.
Large Language Model Supports Multiple Languages
D
mlx-community
789
1
Medgemma 27b Text It Bf16
Other
MedGemma-27B-Text-IT-BF16 is a large language model converted based on the MLX format, focusing on text processing tasks in the medical and health field.
Large Language Model
M
mlx-community
146
1
Medgemma 4b It 8bit
Other
MedGemma-4B-IT-8bit is an MLX format conversion version of the Google MedGemma model, focusing on visual language tasks in the medical field.
Image-to-Text Transformers
M
mlx-community
116
1
Medgemma 4b It 4bit
Other
MedGemma-4B-IT-4bit is a vision-language model specifically designed for the medical field, supporting image and text processing, and suitable for tasks such as medical image analysis.
Image-to-Text Transformers
M
mlx-community
196
1
Devstral Small 2505 Bf16
Apache-2.0
Devstral-Small-2505-bf16 is an MLX-format model converted from MistralAI's Devstral-Small-2505 model, supporting multiple language processing tasks.
Large Language Model Supports Multiple Languages
D
mlx-community
207
1
Gemma 3 12b It 4bit DWQ
A 4-bit quantized version of the Gemma 3 12B model, suitable for the MLX framework and supporting efficient text generation tasks.
Large Language Model
G
mlx-community
554
2
Qwen3 4B 4bit DWQ
Apache-2.0
This model is a 4-bit DWQ quantized version of Qwen3-4B, converted to the MLX format for easy text generation using the mlx library.
Large Language Model
Q
mlx-community
517
2
Qwen3 30B A3B 4bit DWQ 05082025
Apache-2.0
This is a 4-bit quantized model converted from Qwen/Qwen3-30B-A3B to MLX format, suitable for text generation tasks.
Large Language Model
Q
mlx-community
240
5
Qwen3 30B A3B 4bit DWQ 0508
Apache-2.0
Qwen3-30B-A3B-4bit-DWQ-0508 is a 4-bit quantized model converted from Qwen/Qwen3-30B-A3B to MLX format, suitable for text generation tasks.
Large Language Model
Q
mlx-community
410
12
Phi 4 Mini Reasoning MLX 4bit
MIT
This is a 4-bit quantized version in MLX format converted from the Microsoft Phi-4-mini-reasoning model, suitable for text generation tasks.
Large Language Model
P
lmstudio-community
72.19k
2
Josiefied Qwen3 1.7B Abliterated V1 4bit
4-bit quantized version based on Qwen3-1.7B, a lightweight large language model optimized for the MLX framework
Large Language Model
J
mlx-community
135
2
Qwen3 1.7B 4bit
Apache-2.0
Qwen3-1.7B-4bit is a 4-bit quantized version of the Tongyi Qianwen 1.7B model, which has been converted to the MLX framework format for efficient operation on Apple Silicon devices.
Large Language Model
Q
mlx-community
11.85k
2
Dia 1.6B 6bit
Apache-2.0
Dia-1.6B-6bit is a 6-bit quantized version based on Dia-1.6B, optimized for the MLX framework as a text-to-speech model.
Speech Synthesis English
D
mlx-community
76
4
Dia 1.6B 4bit
Apache-2.0
Dia-1.6B-4bit is a 4-bit quantized text-to-speech model based on the MLX format, converted from nari-labs/Dia-1.6B.
Speech Synthesis English
D
mlx-community
168
4
Gemma 3 27b It Qat Bf16
Gemma 3 27B IT QAT BF16 is a version of the Gemma series of models released by Google. It has undergone quantization-aware training (QAT) and is converted to the BF16 format, suitable for the MLX framework.
Image-to-Text Transformers
G
mlx-community
178
2
Gemma 3 27b It Qat 8bit
Other
Gemma 3 27B IT QAT 8bit is an MLX-format model converted from Google's Gemma 3 27B model, supporting image-to-text tasks.
Image-to-Text Transformers Other
G
mlx-community
422
2
Smoldocling 256M Preview Mlx Fp16
Apache-2.0
This model is converted from ds4sd/SmolDocling-256M-preview to the MLX format, supporting image-text-to-text tasks.
Image-to-Text Transformers English
S
ahishamm
24
1
Olmo 2 0325 32B Instruct 4bit
Apache-2.0
This is a 4-bit quantized version converted from the allenai/OLMo-2-0325-32B-Instruct model, optimized for the MLX framework and suitable for text generation tasks.
Large Language Model Transformers English
O
mlx-community
270
10
Deepseek R1 Distill Qwen 32B 4bit
This is the MLX 4-bit quantized version of the DeepSeek-R1-Distill-Qwen-32B model, designed for efficient inference on Apple silicon devices
Large Language Model Transformers
D
mlx-community
130.79k
40
Smollm 135M 4bit
Apache-2.0
This is a 4-bit quantized 135M parameter small language model, suitable for text generation tasks in resource-constrained environments.
Large Language Model Transformers English
S
mlx-community
312
1
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase